← Back to Contents
Note: This page's design and presentation have been enhanced using Claude (Anthropic's AI assistant) to improve visual quality and educational experience.
Week 3 • Sub-Lesson 1

The Cost of Every Prompt

Energy, water, and the environmental footprint of AI at the level of individual interactions

What We'll Cover

Every time you send a message to an AI assistant, something physical happens: electricity flows through data-centre hardware on the other side of the world. This session puts numbers on that — not to make you feel guilty about using AI, but to help you think clearly about its environmental implications as a researcher and as a citizen.

We will look at energy consumption from the level of a single query up to the scale of global AI usage, at water — an environmental cost that receives less attention than carbon — and at why the numbers you read in the media should always be treated with careful scepticism.

This is a topic where the data are genuinely uncertain and where companies have strong incentives to present their products in the best possible light. Part of our job here is to understand why the numbers are uncertain, not just what they are.

⚠️ A Note on Numbers Throughout This Week

Almost all figures in this area come from one of three sources: company self-reports, independent academic estimates, or media extrapolations from those estimates. These often disagree substantially. Throughout these lessons we will flag where figures come from and what assumptions they rest on. Treat any specific number as an order-of-magnitude guide rather than a precise measurement.

🔍 The Transparency Problem

Before we look at any numbers, it is worth understanding why getting accurate figures is so difficult — because the answer shapes how we interpret everything that follows.

What Companies Don't Disclose

Major AI developers — OpenAI, Google, Meta, Microsoft, Anthropic — do not publish detailed energy or emissions data for individual models. What is typically missing:

  • Energy consumed per query (by model type or size)
  • Data centre locations and their grid carbon intensity
  • Water consumption at specific facilities
  • Hardware lifecycle data (manufacturing emissions)
  • Total annual compute for model training

Some companies publish aggregate sustainability reports, but these cover their entire operations, not AI specifically, and rely on accounting methods that can obscure more than they reveal.

What Researchers Can Estimate

In the absence of direct data, researchers use indirect methods to estimate AI's environmental footprint:

  • Hardware benchmarks: Measure energy use of known GPUs, estimate utilisation rates
  • Open-source proxies: Run open-weight models locally and measure directly
  • Architectural inference: Use known model sizes and FLOPs estimates to project energy
  • Company disclosures: Use reported figures with appropriate scepticism about representativeness

This is why published estimates for the same model can differ by a factor of 10 or more.

📄 Key Reading: The Hidden Costs of AI

Nature (2024): "The Hidden Costs of AI: Impending Energy and Resource Strain" — Nature's news team interviews researchers and industry figures on the difficulty of getting accurate data from companies. A good starting point for understanding the transparency problem.

⚡ Energy per Interaction

Let's start at the level of a single query. How much electricity does it take to generate one response?

📊 Energy Consumption by Task Type

The table below combines company-reported figures and independent measurements from MIT Technology Review (2025). Note the wide ranges — these reflect genuine differences in model size, hardware, and methodology, not just uncertainty.

Task Energy Estimate Everyday Comparison Source / Confidence
Google Search query ~0.3 Wh ~11 seconds of a light bulb (LED) Google; moderate confidence
ChatGPT text prompt (reported) 0.34 Wh ~13 seconds of a light bulb OpenAI blog, June 2025; likely best-case
Gemini text prompt (reported) 0.24 Wh ~9 seconds of TV Google White Paper, 2025; likely best-case
Large model text prompt (independent) 0.1–8 Wh 4 seconds – ~5 minutes of a microwave MIT Technology Review, 2025; measured on open-weight models
AI image generation ~2–5 Wh ~1–3 minutes of a microwave Various; moderate confidence
AI video generation (5 seconds) ~3.4 MJ (~944 Wh) Approximately 1 hour of microwave use MIT Technology Review, 2025; ~700× image cost

Key observation: Text generation and video generation are not in the same ballpark — they differ by roughly three orders of magnitude. Company-reported figures for text prompts are at the lower end of independent measurements, likely reflecting optimised infrastructure not available to all deployments.

📹 Video Generation: A Different Category

The energy cost of generating a 5-second AI video (~944 Wh, per MIT Technology Review 2025) is not a typing error. High-resolution video generation involves running a very large diffusion model many times over — the equivalent of generating hundreds or thousands of images sequentially and combining them. This figure comes from measuring open-source video generation models; proprietary models like Sora may differ.

For context: 944 Wh is roughly what an average South African household uses in an entire day of electricity consumption.

✈️ A Worked Example: The Flight Comparison

A common comparison in media coverage: how does using AI compare to taking a transatlantic flight? Let's work through this carefully — because how you do the calculation matters enormously.

🔢 Step-by-Step Calculation

Reference point: one economy-class transatlantic flight (London → New York, ~5,540 km)

  • CO₂ per passenger (direct emissions only): ~0.5 tonnes
  • Including radiative forcing at altitude (contrails, water vapour): roughly doubles the climate impact
  • Commonly used estimate: ~1 tonne CO₂e per passenger
  • Note: figures from different calculators range from 0.5 to 1.5 tonnes — this is itself uncertain

Carbon per ChatGPT text prompt:

Energy assumption Source CO₂ per prompt
(US grid avg: 0.386 kg/kWh)
Calls to match 1-tonne flight
0.34 Wh OpenAI (reported, 2025) 0.13 g CO₂ ~7.6 million
2 Wh Mid-range independent estimate 0.77 g CO₂ ~1.3 million
8 Wh Large model, independent measurement 3.1 g CO₂ ~323,000

What this tells us: Depending entirely on which figures you use, one transatlantic flight equals somewhere between 300,000 and 7.6 million ChatGPT text queries. This is not a rounding error — it reflects the difference between company-optimised infrastructure and real-world large-model deployments, as well as genuine uncertainty about what "a ChatGPT query" even means in terms of model size and compute.

💡 Why the Range Is the Lesson

The fact that this calculation can produce answers spanning two orders of magnitude is not a failure of the analysis — it is the most important finding. It tells us that glib comparisons ("ChatGPT = X flights per day") depend almost entirely on unstated assumptions, and that the opacity of AI companies makes it impossible to pin down a single honest answer.

As researchers, the appropriate response is not to pick the number that fits our preferred narrative, but to present the range honestly and to push for better disclosure.

📄 Source for the energy figures above

MIT Technology Review (2025): "AI energy usage, climate footprint, big tech" — independent measurements of open-source models providing a useful counterpoint to company-reported figures.

Anthony et al. (2023): "Power Hungry Processing: Watts Driving the Cost of AI Deployment?" — benchmarks inference energy across a wide range of NLP tasks and model types.

💧 The Water Footprint

Energy gets most of the attention, but water is the environmental cost of AI that is least reported and least understood. Data centres are thirsty in two distinct ways.

Direct Water Use: Cooling

Modern data centres generate enormous amounts of heat. The most common solution is evaporative cooling — running water over heat exchangers where it evaporates, carrying heat away.

  • Why water, not air? Evaporative cooling is far more energy-efficient than air cooling for the temperatures involved
  • Why filtered municipal water? Minerals in unfiltered water would corrode and clog sensitive hardware — most large data centres use high-quality drinking water
  • Estimate: A 100 MW data centre may consume the equivalent of ~2,600 households' daily water use (IEA estimate)
  • US total: Data centres directly consume roughly 17.5 billion gallons of water per year — about 0.3% of the US public water supply (Lawrence Berkeley National Laboratory, 2024)

Indirect Water Use: Electricity

Generating electricity also consumes water — at the power plant, not the data centre. This "indirect" water use is often larger than the direct cooling water.

  • Thermal power plants (coal, gas, nuclear) use water for steam generation and cooling
  • This water is often not returned to its source — it is evaporated or discharged at a different temperature
  • A commonly cited estimate (Shaolei Ren, UC Riverside): a ~30-turn ChatGPT conversation = roughly 500 ml of water, of which only 12–13% is direct cooling water; the rest is from electricity generation
  • Geographic sensitivity: Water extracted in arid regions has very different consequences from the same volume extracted in water-abundant areas

⚠️ Treating the 500 ml figure with care

The "500 ml per 30-turn conversation" figure (from Li et al., 2023 / Shaolei Ren) became widely cited in media coverage, often without key caveats: it is an estimate based on assumed data centre locations (US-average), a specific model generation (GPT-3 era), and indirect water attribution methods that are contested. Modern data centres vary widely in water efficiency — some use closed-loop systems that recirculate water rather than evaporating it. Treat this as an order-of-magnitude estimate, not a precise measurement. The original paper acknowledges significant uncertainty.

📄 Key Reading: AI Water Footprint

Li, Yang, Islam, Ren (2023): "Making AI Less Thirsty" — the most-cited quantitative analysis of AI water consumption. Read the paper, not just the media coverage of it.

📊 Putting It at Scale

Individual query costs only matter if we multiply them by usage. Let's consider what the aggregate picture looks like.

🌍 AI Usage at Global Scale (2025)

Metric Figure Source / Note
ChatGPT queries per day ~2.5 billion OpenAI (reported to Axios, 2025)
Organisations using AI 78% of surveyed organisations Stanford AI Index, 2025
US data centre electricity (2023) ~100 TWh/year Lawrence Berkeley National Laboratory, 2024 (tripled since 2014)
Projected AI data centre electricity (2028) 250–400 TWh/year Various analysts; high uncertainty
US household electricity reference ~1,200 TWh/year total US EIA; for comparison purposes

If AI data centres reach 400 TWh by 2028, that would be roughly equivalent to one-third of all US household electricity consumption. These projections carry very high uncertainty — they depend on assumptions about AI adoption rates, hardware efficiency improvements, and grid composition.

💡 Training vs. Inference: Where the Energy Goes

Much early coverage of AI's environmental cost focused on the energy required to train a model — the one-time compute cost of creating GPT-4 or Claude. But for widely-deployed models, inference — running the model to answer queries — can easily exceed training energy over the model's lifetime.

With 2.5 billion queries per day, the cumulative inference cost of ChatGPT dwarfs the one-time training cost within months of deployment. This is why statements like "training GPT-3 emitted X tonnes of CO₂" need to be placed in the context of ongoing inference costs.

📚 Summary & Key Takeaways

Before we can have a productive conversation about AI's environmental impact, we need to understand the numbers — and their limits:

  • Corporate opacity is the fundamental problem: AI companies don't disclose the data needed to calculate their environmental footprint, so all estimates involve assumptions
  • Text vs. video is not comparable: Video generation uses roughly 1,000× more energy than text generation — these are different categories of use
  • The flight comparison spans orders of magnitude: 300,000 to 7.6 million queries per flight, depending on model size and whose figures you trust
  • Water is underreported: Both direct cooling and indirect electricity generation consume significant water, with geographic implications
  • Scale is what matters: Individual query costs are small; billions of queries per day is not
  • Inference, not just training: The ongoing cost of running models at scale quickly exceeds one-time training costs

Next session (Week 3.2): We zoom out from individual queries to the infrastructure level — where does the electricity come from, how does manufacturing hardware fit in, and why do efficiency gains so often fail to reduce total energy use?